Goto

Collaborating Authors

 inverse reinforcement learning


Generative Adversarial Imitation Learning

Neural Information Processing Systems

Consider learning a policy from example expert behavior, without interaction with the expert or access to a reinforcement signal. One approach is to recover the expert's cost function with inverse reinforcement learning, then extract a policy from that cost function with reinforcement learning. This approach is indirect and can be slow. We propose a new general framework for directly extracting a policy from data as if it were obtained by reinforcement learning following inverse reinforcement learning. We show that a certain instantiation of our framework draws an analogy between imitation learning and generative adversarial networks, from which we derive a model-free imitation learning algorithm that obtains significant performance gains over existing model-free methods in imitating complex behaviors in large, high-dimensional environments.








Bayesian Inference of Temporal Task Specifications from Demonstrations

Ankit Shah, Pritish Kamath, Julie A. Shah, Shen Li

Neural Information Processing Systems

Temporal logics have been used in prior research as a language forexpressing desirable system behaviors, and canimprovetheinterpretability ofspecifications if expressed as compositions of simpler templates (akin to those described by Dwyer et al. [2]).